Comparative study of boosting and non-boosting training for constructing ensembles of acoustic models
نویسندگان
چکیده
This paper compares the performance of Boosting and nonBoosting training algorithms in large vocabulary continuous speech recognition (LVCSR) using ensembles of acoustic models. Both algorithms demonstrated significant word error rate reduction on the CMU Communicator corpus. However, both algorithms produced comparable improvements, even though one would expect that the Boosting algorithm, which has a solid theoretic foundation, should work much better than the non-Boosting algorithm. Several voting schemes for hypothesis combining were evaluated, including weighted voting, un-weighted voting and ROVER.
منابع مشابه
Improving the performance of an LVCSR system through ensembles of acoustic models
This paper describes our work on applying ensembles of acoustic models to the problem of large vocabulary continuous speech recognition (LVCSR). We propose three algorithms for constructing ensembles. The first two have their roots in bagging algorithms; however, instead of randomly sampling examples our algorithms construct training sets based on the word error rate. The third one is a boostin...
متن کاملData selection and model combination in connectionist speech recognition
The hybrid connectionist-hidden Markov model (HMM) approach to large vocabulary continuous speech recognition has been shown to be competitive with HMM based systems. However, the recent availability of extremely large amounts of acoustic training data has highlighted a problem with the connectionist acoustic modelling paradigm. The effective use of such large amounts of data is difficult due t...
متن کاملOptimizing boosting with discriminative criteria
We describe the use of discriminative criteria to optimize Boosting based ensembles. Boosting algorithms may create hundreds of individual classifiers in order to fit the training data. However, this strategy isn’t feasible and necessary for complex classification problems, such as real-time continuous speech recognition, in which only the combination of a few of acoustic models is practical. H...
متن کاملA Case Study on Bagging, Boosting, and Basic Ensembles of Neural Networks for OCR
W e study the effectiveness of three neural network ensembles in improving OCR performance: ( i ) Basic, (ii) Bagging, and (iii) Boosting. Three random character degradation models are introduced in training indivadual networks in order to reduce error correlation between individual networks and to improve the generalization ability of neural networks. We compare the recognition accuracies of t...
متن کاملConstructing Diverse Classifier Ensembles using Artificial Training Examples
Ensemble methods like bagging and boosting that combine the decisions of multiple hypotheses are some of the strongest existing machine learning methods. The diversity of the members of an ensemble is known to be an important factor in determining its generalization error. This paper presents a new method for generating ensembles that directly constructs diverse hypotheses using additional arti...
متن کامل